67 research outputs found

    Attitude Recognition Using Multi-resolution Cochleagram Features

    Get PDF

    User Identity Protection in Automatic Emotion Recognition through Disguised Speech

    Get PDF
    Ambient Assisted Living (AAL) technologies are being developed which could assist elderly people to live healthy and active lives. These technologies have been used to monitor people’s daily exercises, consumption of calories and sleep patterns, and to provide coaching interventions to foster positive behaviour. Speech and audio processing can be used to complement such AAL technologies to inform interventions for healthy ageing by analyzing speech data captured in the user’s home. However, collection of data in home settings presents challenges. One of the most pressing challenges concerns how to manage privacy and data protection. To address this issue, we proposed a low cost system for recording disguised speech signals which can protect user identity by using pitch shifting. The disguised speech so recorded can then be used for training machine learning models for affective behaviour monitoring. Affective behaviour could provide an indicator of the onset of mental health issues such as depression and cognitive impairment, and help develop clinical tools for automatically detecting and monitoring disease progression. In this article, acoustic features extracted from the non-disguised and disguised speech are evaluated in an affect recognition task using six different machine learning classification methods. The results of transfer learning from non-disguised to disguised speech are also demonstrated. We have identified sets of acoustic features which are not affected by the pitch shifting algorithm and also evaluated them in affect recognition. We found that, while the non-disguised speech signal gives the best Unweighted Average Recall (UAR) of 80.01%, the disguised speech signal only causes a slight degradation of performance, reaching 76.29%. The transfer learning from non-disguised to disguised speech results in a reduction of UAR (65.13%). However, feature selection improves the UAR (68.32%). This approach forms part of a large project which includes health and wellbeing monitoring and coaching

    Machine Learning models for detection and assessment of progression in Alzheimer's disease based on blood and cerebrospinal fluid biomarkers

    Get PDF
    Machine-learning techniques were applied to human blood plasma and cerebrospinal fluid (CSF) biomarker data related to cognitive decline in Alzheimer’s Disease (AD) patients available via Alzheimer Disease Neuroimaging Initiative (ADNI) study. We observed the accuracy of AD diagnosis is greatest when protein biomarkers from cerebrospinal fluid are combined with plasma proteins using Support Vector Machines (SVM); this is not improved by adding age and sex. The area under the receiver operator characteristic (ROC) curve for our model of AD diagnosis based on a full (unbiased) set of plasma proteins was 0.94 in cross-validation and 0.82 on an external validation (test) set. Taking plasma in combination with CSF, the model reaches 0.98 area under the ROC curve on the test set. Accuracy of prediction of risk of mild cognitive impairment progressing to AD is the same for blood plasma biomarkers as for CSF and is not improved by combining them or adding age and sex as covariates.Clinical relevance— The identification of accurate and cost-effective biomarkers to screen for risk of developing AD and monitoring its progression is crucial for improved understanding of its causes and stratification of patients for treatments under development. This paper demonstrates the feasibility of AD detection and prognosis based on blood plasma biomarkers.<br/

    Temporal Integration of Text Transcripts and Acoustic Features for Alzheimer's Diagnosis Based on Spontaneous Speech

    Get PDF
    Background: Advances in machine learning (ML) technology have opened new avenues for detection and monitoring of cognitive decline. In this study, a multimodal approach to Alzheimer's dementia detection based on the patient's spontaneous speech is presented. This approach was tested on a standard, publicly available Alzheimer's speech dataset for comparability. The data comprise voice samples from 156 participants (1:1 ratio of Alzheimer's to control), matched by age and gender. Materials and Methods: A recently developed Active Data Representation (ADR) technique for voice processing was employed as a framework for fusion of acoustic and textual features at sentence and word level. Temporal aspects of textual features were investigated in conjunction with acoustic features in order to shed light on the temporal interplay between paralinguistic (acoustic) and linguistic (textual) aspects of Alzheimer's speech. Combinations between several configurations of ADR features and more traditional bag-of-n-grams approaches were used in an ensemble of classifiers built and evaluated on a standardised dataset containing recorded speech of scene descriptions and textual transcripts. Results: Employing only semantic bag-of-n-grams features, an accuracy of 89.58% was achieved in distinguishing between Alzheimer's patients and healthy controls. Adding temporal and structural information by combining bag-of-n-grams features with ADR audio/textual features, the accuracy could be improved to 91.67% on the test set. An accuracy of 93.75% was achieved through late fusion of the three best feature configurations, which corresponds to a 4.7% improvement over the best result reported in the literature for this dataset. Conclusion: The proposed combination of ADR audio and textual features is capable of successfully modelling temporal aspects of the data. The machine learning approach toward dementia detection achieves best performance when ADR features are combined with strong semantic bag-of-n-grams features. This combination leads to state-of-the-art performance on the AD classification task

    Extracting audio-visual features for emotion recognition through active feature selection

    Get PDF
    • …
    corecore